Good old on-line back-propagation for plain multi-layer perceptrons yields avery low 0.35% error rate on the famous MNIST handwritten digits benchmark. Allwe need to achieve this best result so far are many hidden layers, many neuronsper layer, numerous deformed training images, and graphics cards to greatlyspeed up learning.
展开▼